36 research outputs found

    Trimming of mammalian transcriptional networks using network component analysis

    Get PDF
    <p>Abstract</p> <p>Background</p> <p>Network Component Analysis (NCA) has been used to deduce the activities of transcription factors (TFs) from gene expression data and the TF-gene binding relationship. However, the TF-gene interaction varies in different environmental conditions and tissues, but such information is rarely available and cannot be predicted simply by motif analysis. Thus, it is beneficial to identify key TF-gene interactions under the experimental condition based on transcriptome data. Such information would be useful in identifying key regulatory pathways and gene markers of TFs in further studies.</p> <p>Results</p> <p>We developed an algorithm to trim network connectivity such that the important regulatory interactions between the TFs and the genes were retained and the regulatory signals were deduced. Theoretical studies demonstrated that the regulatory signals were accurately reconstructed even in the case where only three independent transcriptome datasets were available. At least 80% of the main target genes were correctly predicted in the extreme condition of high noise level and small number of datasets. Our algorithm was tested with transcriptome data taken from mice under rapamycin treatment. The initial network topology from the literature contains 70 TFs, 778 genes, and 1423 edges between the TFs and genes. Our method retained 1074 edges (i.e. 75% of the original edge number) and identified 17 TFs as being significantly perturbed under the experimental condition. Twelve of these TFs are involved in MAPK signaling or myeloid leukemia pathways defined in the KEGG database, or are known to physically interact with each other. Additionally, four of these TFs, which are Hif1a, Cebpb, Nfkb1, and Atf1, are known targets of rapamycin. Furthermore, the trimmed network was able to predict <it>Eno1 </it>as an important target of Hif1a; this key interaction could not be detected without trimming the regulatory network.</p> <p>Conclusions</p> <p>The advantage of our new algorithm, relative to the original NCA, is that our algorithm can identify the important TF-gene interactions. Identifying the important TF-gene interactions is crucial for understanding the roles of pleiotropic global regulators, such as p53. Also, our algorithm has been developed to overcome NCA's inability to analyze large networks where multiple TFs regulate a single gene. Thus, our algorithm extends the applicability of NCA to the realm of mammalian regulatory network analysis.</p

    Low-density series expansions for directed percolation I: A new efficient algorithm with applications to the square lattice

    Full text link
    A new algorithm for the derivation of low-density series for percolation on directed lattices is introduced and applied to the square lattice bond and site problems. Numerical evidence shows that the computational complexity grows exponentially, but with a growth factor \lambda < \protect{\sqrt[8]{2}}, which is much smaller than the growth factor \lambda = \protect{\sqrt[4]{2}} of the previous best algorithm. For bond (site) percolation on the directed square lattice the series has been extended to order 171 (158). Analysis of the series yields sharper estimates of the critical points and exponents.Comment: 20 pages, 8 figures (3 of them > 1Mb

    Technologies and Approaches to Elucidate and Model the Virulence Program of Salmonella

    Get PDF
    Salmonella is a primary cause of enteric diseases in a variety of animals. During its evolution into a pathogenic bacterium, Salmonella acquired an elaborate regulatory network that responds to multiple environmental stimuli within host animals and integrates them resulting in fine regulation of the virulence program. The coordinated action by this regulatory network involves numerous virulence regulators, necessitating genome-wide profiling analysis to assess and combine efforts from multiple regulons. In this review we discuss recent high-throughput analytic approaches used to understand the regulatory network of Salmonella that controls virulence processes. Application of high-throughput analyses have generated large amounts of data and necessitated the development of computational approaches for data integration. Therefore, we also cover computer-aided network analyses to infer regulatory networks, and demonstrate how genome-scale data can be used to construct regulatory and metabolic systems models of Salmonella pathogenesis. Genes that are coordinately controlled by multiple virulence regulators under infectious conditions are more likely to be important for pathogenesis. Thus, reconstructing the global regulatory network during infection or, at the very least, under conditions that mimic the host cellular environment not only provides a bird's eye view of Salmonella survival strategy in response to hostile host environments but also serves as an efficient means to identify novel virulence factors that are essential for Salmonella to accomplish systemic infection in the host

    A community effort towards a knowledge-base and mathematical model of the human pathogen Salmonella Typhimurium LT2

    Get PDF
    <p>Abstract</p> <p>Background</p> <p>Metabolic reconstructions (MRs) are common denominators in systems biology and represent biochemical, genetic, and genomic (BiGG) knowledge-bases for target organisms by capturing currently available information in a consistent, structured manner. <it>Salmonella enterica </it>subspecies I serovar Typhimurium is a human pathogen, causes various diseases and its increasing antibiotic resistance poses a public health problem.</p> <p>Results</p> <p>Here, we describe a community-driven effort, in which more than 20 experts in <it>S</it>. Typhimurium biology and systems biology collaborated to reconcile and expand the <it>S</it>. Typhimurium BiGG knowledge-base. The consensus MR was obtained starting from two independently developed MRs for <it>S</it>. Typhimurium. Key results of this reconstruction jamboree include i) development and implementation of a community-based workflow for MR annotation and reconciliation; ii) incorporation of thermodynamic information; and iii) use of the consensus MR to identify potential multi-target drug therapy approaches.</p> <p>Conclusion</p> <p>Taken together, with the growing number of parallel MRs a structured, community-driven approach will be necessary to maximize quality while increasing adoption of MRs in experimental design and interpretation.</p

    SBML Level 3: an extensible format for the exchange and reuse of biological models

    Get PDF
    Systems biology has experienced dramatic growth in the number, size, and complexity of computational models. To reproduce simulation results and reuse models, researchers must exchange unambiguous model descriptions. We review the latest edition of the Systems Biology Markup Language (SBML), a format designed for this purpose. A community of modelers and software authors developed SBML Level 3 over the past decade. Its modular form consists of a core suited to representing reaction-based models and packages that extend the core with features suited to other model types including constraint-based models, reaction-diffusion models, logical network models, and rule-based models. The format leverages two decades of SBML and a rich software ecosystem that transformed how systems biologists build and interact with models. More recently, the rise of multiscale models of whole cells and organs, and new data sources such as single-cell measurements and live imaging, has precipitated new ways of integrating data with models. We provide our perspectives on the challenges presented by these developments and how SBML Level 3 provides the foundation needed to support this evolution

    Data Mining Based on Principal Component Analysis: Application to the Nitric Oxide Response in Escherichia coli

    No full text
    This work evaluates a recently developed multivariate statistical method based on the creation of pseudo or latent variables using principal component analysis (PCA). The application is the data mining of gene expression data to find a small subset of the most important genes in a set of thousand or tens of thousands of genes from a relatively small number of experimental runs. The method was previously developed and evaluated on artificially generated data and real data sets. Its evaluations consisted of its ability to rank the genes against known truth in simulated data studies and to identify known important genes in real data studies. The purpose of the work described here is to identify a ranked set of genes in an experimental study and then for a few of the most highly ranked unverified genes, experimentally verify their importance.This method was evaluated using the transcriptional response of Escherichia coli to treatment with four distinct inhibitory compounds: nitric oxide, S-nitrosoglutathione, serine hydroxamate and potassium cyanide. Our analysis identified genes previously recognized in the response to these compounds and also identified new genes.Three of these new genes, ycbR, yfhA and yahN, were found to significantly (p-values<0.002) affect the sensitivityof E. coli to nitric oxide-mediated growth inhibition. Given that the three genes were not highly ranked in the selected ranked set (RS), these results support strong sensitivity in the ability of the method to successfully identify genes related to challenge by NO and GSNO. This ability to identify genes related to the response to an inhibitory compound is important for engineering tolerance to inhibitory metabolic products, such as biofuels, and utilization of cheap sugar streams, such as biomass-derived sugars or hydrolysate.This article is published as Teh, AiLing, D. S. Layton, Daniel R. Hyduke, Laura R. Jarboe, and D. K. Rollins. "Data Mining Based on Principal Component Analysis: Application to the Nitric Oxide Response in Escherichia coli." Journal of Statistical Science and Application 2 (2014): 1-18. DOI: 10.17265/2328-224X/2014.01.001. Attribution-NonCommercial 4.0 International (CC BY-NC 4.0). Copyright 2014 David Publishing Company. Posted with permission
    corecore